Theory and Practice of Expectation Maximization (EM) Algorithm
نویسندگان
چکیده
In the field of statistical data mining, the Expectation Maximization (EM) algorithm is one of the most popular methods used for solving parameter estimation problems in the maximum likelihood (ML) framework. Compared to traditional methods such as steepest descent, conjugate gradient, or Newton-Raphson, which are often too complicated to use in solving these problems, EM has become a popular method because it takes advantage of some problem specific properties (Xu et al., 1996). The EM algorithm converges to the local maximum of the log-likelihood function under very general conditions (Demspter et al., 1977; Redner et al., 1984). Efficiently maximizing the likelihood by augmenting it with latent variables and guarantees of convergence are some of the important hallmarks of the EM algorithm. EM based methods have been applied successfully to solve a wide range of problems that arise in fields of pattern recognition, clustering, information retrieval, computer vision, bioinformatics (Reddy et al., 2006; Carson et al., 2002; Nigam et al., 2000), etc. Given an initial set of parameters, the EM algorithm can be implemented to compute parameter estimates that locally maximize the likelihood function of the data. In spite of its strong theoretical foundations, its wide applicability and important usage in solving some real-world problems, the standard EM algorithm suffers from certain fundamental drawbacks when used in practical settings. Some of the main difficulties of using the EM algorithm on a general log-likelihood surface are as follows (Reddy et al., 2008):
منابع مشابه
Expectation Maximization
The Expectation Maximization (EM) algorithm [1, 2] is one of the most widely used algorithms in statistics. Suppose we are given some observed data X and a model family parametrized by θ, and would like to find the θ which maximizes p(X |θ), i.e. the maximum likelihood estimator. The basic idea of EM is actually quite simple: when direct maximization of p(X |θ) is complicated we can augment the...
متن کاملNoisy Expectation-Maximization: Applications and Generalizations
We present a noise-injected version of the Expectation-Maximization (EM) algorithm: the Noisy Expectation Maximization (NEM) algorithm. The NEM algorithm uses noise to speed up the convergence of the EM algorithm. The NEM theorem shows that injected noise speeds up the average convergence of the EM algorithm to a local maximum of the likelihood surface if a positivity condition holds. The gener...
متن کاملAn Explanation of the Expectation Maximization Algorithm, Report no. LiTH-ISY-R-2915
The expectation maximization (EM) algorithm computes maximum likelihood estimates of unknown parameters in probabilistic models involving latent variables. More pragmatically speaking, the EM algorithm is an iterative method that alternates between computing a conditional expectation and solving a maximization problem, hence the name expectation maximization. We will in this work derive the EM ...
متن کاملSpectrometer Calibration by Expectation Maximization Method
Expectation Maximization (EM) algorithm is a parameter estimation method from incomplete observations. In this paper, an implementation of this method to the calibration of HKS spectrometer at Jefferson Lab is described. We show that the application of EM method is able to calibrate the spectrometer properly in the presence of high background noise, while the traditional nonlinear Least Square ...
متن کاملQuantitative SPECT and planar 32P bremsstrahlung imaging for dosimetry purpose –An experimental phantom study
Background: In this study, Quantitative 32P bremsstrahlung planar and SPECT imaging and consequent dose assessment were carried out as a comprehensive phantom study to define an appropriate method for accurate Dosimetry in clinical practice. Materials and Methods: CT, planar and SPECT bremsstrahlung images of Jaszczak phantom containing a known activity of 32P were acquired. In addition, Phanto...
متن کامل